Snowflake Custom Transformation Job

Data Pipeline Studio (DPS) provides templates for creating transformation jobs. The jobs include Join/ Union/ Aggregate functions that can be performed to group or combine data for analysis.

For complex operations to be performed on data, Lazsa DPS provides the option of creating custom transformation jobs. For custom queries while the logic is written by the users, DPS UI provides an option to create SQL queries by selecting specific columns of tables. Lazsa consumes the SQL queries along with the transformation logic, to generate the code for custom transformation jobs.

To create a custom transformation job

  1. Sign in to the Lazsa Platform and navigate to Products.

  2. Select a product and feature. Click the Develop stage of the feature, you are navigated to Data Pipeline Studio.

  3. Create a pipeline with the following nodes:

    Note: The stages and technologies used in this pipeline are merely for the sake of example.

    • Data Lake - Snowflake

    • Data Transformation - Snowflake

      Snowflake Transformation Pipeline

  4. Click the data lake node to configure it. You can do one of the following:

    • Click Configured Datastore and select a datastore from the dropdown list. The options in the dropdown list are seen only if you have access to one or more configured Snowflake accounts. Select the Warehouse, Database, and Schema from the dropdown options.

    • Click New Datastore. See Create New Datastore

  5. Click the data transformation node, and then click Create Custom Job. Provide the following information:

Snowflake Custom Transformation Job

    This way you can add multiple queries. Once you are done adding queries, click Complete. A message appears that the Snowflake custom transformation job is created.

  1. Click Custom Script. Do one of the following:

    • If you have the custom script file ready, click Browse this computer and drop the file.

    • If you want to create the custom script file, click Download Template. Create the file and then upload it.

  2. Click Complete.

  3. On home page of DPS, click Publish to publish the pipeline.

  1. Click the Snowflake Transformation stage and click Start to start the job run.

Start job run

After the job run is complete how do I view the details?

There are two ways to view the details:

  1. Click on the Snowflake Data Lake node. Click the ellipsis (...) and click Dashboard. You are navigated to the Snowflake dashboard, where you can view the job run details.

    View job details

  1. Browse to the table in Data Browsing and view the records from the file.

    Data browsing

Related Topics Link IconRecommended Topics

What's next? Databricks Custom Transformation Job